After I hit start training in about 10 seconds it says "Training Complete. Check the outputs folder for the LoRA files." Could it be because I have multiple GPUs in my system? I noticed in the train log it does say something about multi_gpu. Not an error. But thought maybe it's not supported. My GPU's aren't linked or anything.
@@ChrissyAiven jeje that's the issue it just gave me a rare file I don't remember what was the extension of the file or if it was an orphan file I call it that way jeje to those files without an extension like .jpg etc. Also the the size of the file was just a few KB
My hit rate on these tutorials has been pretty low. There's always something that doesn't work no matter what you do (reactor 🤨). This tutorial however, not only worked, I was amazed at the results. This really is a game changer. I can now make my own Loras and that is fking cool! Thank you. PS. Although it's trained on Flux1-dev, the resulting Lora works great on Q5_K_S.
thank you! works! one detail. depending the python , probably some conflicts will happen. the setup on the page right now (nov 2024) works with 3.10 (64 bits) in my computer (AMD)
FileNotFoundError: [WinError 3] The system cannot find the path specified: 'C:\\Users\\ishwr\\Music\\fluxgym\\outputs\\lora1\\sample' still tried alot but getting same error
Great video. I tried to set it up on my laptop but it didn't work. Can you create a Colab Notebook for this? I tried to do one but having trouble setting up the WebUI hosting part. It points to a local host.
@@aidendeans5569 I got some replies on Reddit people said that it's working, the not reporting progress thing is a bug that the FluxGym dev gave up on I guess. But its still working correctly. I myself switched to using Kohya_ss with its GUI directly instead (using the SD3_Flux.1 branch of Kohya) and its working for me
Never mind I read in comments you said you haven’t tried with other quantizations. I have a 4080 (16gb vram) but it should like I should try the standard flux 1 dev model safetensor
for some reason after i press ADD AI CAPTION WITH FELORNACE 2 my cmd stuck to downloading: pytorch_model.bin how can i fix this? is there anyway to reject downloading this? i mean is there anywat to download it manulay and replace it ? where should i download and where should i place it?
Hi my fluxgym is running for around 12 hours now, I have a 16GB 4080, so should I stop it? The sample files showing at the bottom are really strange too, nothing to do with the model or is it normal?
@@RonnieMirands hm maybe ... btw I have a 4060 got the number wrong. Trying it on Flux Trainer inside ComfyUI now but seems the same, GPU is on 100% estimated time 15 hours. Well I don't mind IF it is working after this, but it seems to be wrong anyways :D
@@ChrissyAiven I´ve search more about this, and seems it really take a lot of hours! Thats why in beggining they talked it was possible train just with high end cards. So they optmized a lot, but still take lot of time :(
@@RonnieMirands oh btw did you try this version and also the trainer within comfy? Gonna make a lora on each and then compare. My face is from fooocus and I wasn't able to rebuild it on comfy, so I hope Lora works :)
SAME HERE been trying to get it working for 2 days now and it only can just keep downloading giant model files on my drive and doesn't ever start even training anything haha
@@neolamas147 I give up on this, my last hope just run it on google colab, but I have to upgrade to colab pro its 9.99, because standard colab have limited time
Do you know how to traine using multiple datasets, Like in buked? I would want to create a LoRa using 512, 768 and 1024 at the same time, but I don't know how to do it (more precisely, where to store the aditional images, since the option of buked is avaiable in the advanced tab).
The good news is Flux lora training are able to take multiple size image in 1 dataset. So you can put all different sizes image into 1 training dataset. :)
one question; It seems that if i train a lora with a model and use it with another model, it does not perform fine. Is it possible in Fluxgym to set a different train model ? Like editing any settings file ?
Fluxgym is underrated, only thing that confused me at first was how in the instructions it said to diwnload some flux-dev.sft file instead of the isual safetensor extension .. i simply copied my already existing safetensor in the fluxgym model unet folder and renamed it to have a sft extendion , did the same with the vao and it worked , any thoughts?
@@TheFutureThinker thank you , I was thinking it could stand for supervised fine tuning or something 😅 also another "rule" I broke was using the fp8 T5 text encoder during training by renaming it to fp16 and the scripts seem to correctly recognize it as fp8! Got any experience with using fp8 over fp16?
Same here.... Says Training Complete. Check the outputs folder for the LoRA files after about 10 seconds. Did you figure it out. By chance do you have multiple GPU's in your system?
@@brianmonarchcomedy no. I have one rtx 4070ti. I just let it be. Just wait for another tool. Like kohya ss. Maybe those guys make it right. Or something else.
@@NGIgri I did notice that they might have come up with a new version since this video was made. I wonder if there's a glitch in the new version. You can see that there's a couple differences. Like on my version you type in the resolution you want. But in this version in the video, you have to check off 512 or 1024.
@@brianmonarchcomedy yeah. And I found strange the resolution changes itself. I tried to train LoRA on 1024, but it automaticaly changes to 512. Maybe it's because of12gb of VRAM.
I have some errors trying to figure out but not sure if anybody had the same [INFO] RuntimeError: use_libuv was requested but PyTorch was build without libuv support [ERROR] Command exited with code 1. for some reason doesn't see libuv did the same in other slower machine and works no idea what to do I try different version of pytorch and same.
I found the solution, basically is because I have two GPUs, so the script is trying to use both RTX3090 and for this reason, doesn't work. I just added the extra calls " --num_processes=1 ^ --num_machines=1 ^" on the "train.bat" file and now is using only one GPU without any error. I guess there must be something extra to be installed for multiple GPUs to work. Thanks again for the tutorials was very usefull
@@jocg9168 Having same issues. How do you add those extra calls to the train script? I edit and then it disappears? Train.bat location? Not seeing that. Thanks for this info as it's definitely not with the pytorch etc. I'm a newb at this part.
are you using 1024 image ? ive tried this and comfyui flux training, the comfyui give me much better result (likeliness) than the flux gym, have you tried to compare lora from this and comfyui ? nice simple vid tho, gonna try it latter using your setting, and it seems like thee size of the dataset is also crucial ? (size as in bytes per image not in how many files)
@@TheFutureThinker I have 8GB VRAM and 16 RAM, but I instantly get out of memory with that default models. 😕 Pinokio either crashes completely or I get an error in the terminal
Has the new kohya_ss update tutorial come so that it can work with the FLUX model? I recently installed it, but even though it was the latest update, it did not support FLUX. If possible, teach us how to update it. thanks
Kinda crazy how just vram adds 3x faster speed. I just wish nvidia sold a 48g. Vram graphics card at a 4070ti price point. It costs them nothing to double vram. And given the price of 4070tj we got ripped off hard with 12gb vram. So sad. Even games in 4k you benefit from 24gb vram. They just want to upsell. Given the price of a 4090 it should be 48gb vram too
The character is a very simple example, I suppose transferring the artist’s style will require 200-300 initial scans, and the training time for 12-16 GB will increase to 72 hours?
I started a character lora on my 12Gb VRAM card. It was 25 images at 1024, 1500 steps and I think everything else was default. Once it got to the epoch part (which took awhile), the estimated time was 23hrs. I didn't continue but I will try again with less images, steps and epochs.
@@TheFutureThinker Roop almost runs on anything. They had fast SD CPU, but the trend went to high VRAM then now lower VRAM. There was the LCM LoRA models with Fast SD CPU. Most of this is just photoshop on steroids, but there should be more of a manual way of doing it locally for almost anyone that does not have the VRAM. IMO.
@@TheFutureThinker tried a couple, no fine tuning of course. It was alright for charactera but did not try entire styles, quess the key words, having the text files and training it with more detail might have better results.
Once I hit training, at the bottom I'm seeing : [INFO] RuntimeError: use_libuv was requested but PyTorch was build without libuv support [ERROR] Command exited with code 1 [INFO] Runner: Out of my element, like Donny, on this one. Do I have to install it with this? pip install torch --no-cache-dir If anyone has some insight. Would appreciate it. Everything works perfect until this point.
After I hit start training in about 10 seconds it says "Training Complete. Check the outputs folder for the LoRA files." Could it be because I have multiple GPUs in my system? I noticed in the train log it does say something about multi_gpu. Not an error. But thought maybe it's not supported. My GPU's aren't linked or anything.
A few seconds? And did the lora work? If so..gimme your computer 😁
@@ChrissyAiven jeje that's the issue it just gave me a rare file I don't remember what was the extension of the file or if it was an orphan file I call it that way jeje to those files without an extension like .jpg etc. Also the the size of the file was just a few KB
I was looking for affordable free lora training. God bless you !
Thanks, glad it help
if you issue the command python3.10 venv venv, to activate the environment, it helps to resolve any incompatibility issues.
My hit rate on these tutorials has been pretty low. There's always something that doesn't work no matter what you do (reactor 🤨).
This tutorial however, not only worked, I was amazed at the results. This really is a game changer. I can now make my own Loras and that is fking cool! Thank you.
PS. Although it's trained on Flux1-dev, the resulting Lora works great on Q5_K_S.
Nice 👍 have fun with it.
Yes Flux 1Dev Train lora could work on other flux gguf quantization version.
thank you! works! one detail. depending the python , probably some conflicts will happen. the setup on the page right now (nov 2024) works with 3.10 (64 bits) in my computer (AMD)
Hello, what is your default py reader? I dont usually have a preview of py files on the right when browsing with windows explorer, for example at 6:40
i just got complete the task in a few sec and have a train lora bat file? in the output folder. i dont get it.
FileNotFoundError: [WinError 3] The system cannot find the path specified: 'C:\\Users\\ishwr\\Music\\fluxgym\\outputs\\lora1\\sample'
still tried alot but getting same error
Nice job. Easy explanation. Precise . Great results on lora . Thanx
You are welcome
Great video. I tried to set it up on my laptop but it didn't work. Can you create a Colab Notebook for this? I tried to do one but having trouble setting up the WebUI hosting part. It points to a local host.
Thank you for the amazing tutorial bro
Is it possible to use a T5 fp8 instead of the T5 fp16 model for faster training?
I get an error after 10 seconds of training where it seems that the GPU is not configured. Can someone help me?
@@TiagoTorressoria same here its run about 10 min, it finished early, when i cheked process it said error code 1
I just get stuck where it says it's creating epoch 1/16 and it just sits there, no progress reporting at all
same problem
yeah i same the same issue
@@aidendeans5569 I got some replies on Reddit people said that it's working, the not reporting progress thing is a bug that the FluxGym dev gave up on I guess. But its still working correctly. I myself switched to using Kohya_ss with its GUI directly instead (using the SD3_Flux.1 branch of Kohya) and its working for me
Image preparation is also very important and you did not mention it in the video above.
Prepare metadata, number of images?
Can you train using the flux NF4 safetensor model?
Never mind I read in comments you said you haven’t tried with other quantizations. I have a 4080 (16gb vram) but it should like I should try the standard flux 1 dev model safetensor
FileNotFoundError: [WinError 3] The system cannot find the path specified, i got this error anyone can help please?
i have installed this 6 times and still get this error RuntimeError: use_libuv was requested but PyTorch was build without libuv support
for some reason after i press ADD AI CAPTION WITH FELORNACE 2 my cmd stuck to downloading: pytorch_model.bin how can i fix this? is there anyway to reject downloading this? i mean is there anywat to download it manulay and replace it ? where should i download and where should i place it?
Hi my fluxgym is running for around 12 hours now, I have a 16GB 4080, so should I stop it? The sample files showing at the bottom are really strange too, nothing to do with the model or is it normal?
Mine is running about 4 hours, i thought it was just me lol. So there is something wrong maybe
@@RonnieMirands hm maybe ... btw I have a 4060 got the number wrong. Trying it on Flux Trainer inside ComfyUI now but seems the same, GPU is on 100% estimated time 15 hours.
Well I don't mind IF it is working after this, but it seems to be wrong anyways :D
@@ChrissyAiven I´ve search more about this, and seems it really take a lot of hours! Thats why in beggining they talked it was possible train just with high end cards. So they optmized a lot, but still take lot of time :(
@@RonnieMirands ok thank you, then I will wait patently, not gonna make a lora everyday right? :)
@@RonnieMirands oh btw did you try this version and also the trainer within comfy? Gonna make a lora on each and then compare. My face is from fooocus and I wasn't able to rebuild it on comfy, so I hope Lora works :)
In my case lora file is not generating after whole process , I have 32 gb ram and 12 vrm gpu
SAME HERE been trying to get it working for 2 days now and it only can just keep downloading giant model files on my drive and doesn't ever start even training anything haha
@@neolamas147 I give up on this, my last hope just run it on google colab, but I have to upgrade to colab pro its 9.99, because standard colab have limited time
Do you know how to traine using multiple datasets, Like in buked? I would want to create a LoRa using 512, 768 and 1024 at the same time, but I don't know how to do it (more precisely, where to store the aditional images, since the option of buked is avaiable in the advanced tab).
The good news is Flux lora training are able to take multiple size image in 1 dataset. So you can put all different sizes image into 1 training dataset. :)
can this lora works for SDXL as well?
Great tut. I have a question, How do we resume an interrupted training from the last state?
one question; It seems that if i train a lora with a model and use it with another model, it does not perform fine. Is it possible in Fluxgym to set a different train model ? Like editing any settings file ?
From what I know you can only use them on flux, for SDXL or so you need a different trainer.
Thank you so much for your effort, having difficulty in part two (Add AI caption with Florence 2. Could not sort it out.
Fluxgym is underrated, only thing that confused me at first was how in the instructions it said to diwnload some flux-dev.sft file instead of the isual safetensor extension .. i simply copied my already existing safetensor in the fluxgym model unet folder and renamed it to have a sft extendion , did the same with the vao and it worked , any thoughts?
Sft, it is short terms for safetensors. Both works. I have my VAE named .sft also.
@@TheFutureThinker thank you , I was thinking it could stand for supervised fine tuning or something 😅 also another "rule" I broke was using the fp8 T5 text encoder during training by renaming it to fp16 and the scripts seem to correctly recognize it as fp8! Got any experience with using fp8 over fp16?
Well... This doesn't work fo rme. It ends training right after I started it. No loras in output folder.
Same here.... Says Training Complete. Check the outputs folder for the LoRA files after about 10 seconds. Did you figure it out. By chance do you have multiple GPU's in your system?
@@brianmonarchcomedy no. I have one rtx 4070ti. I just let it be. Just wait for another tool. Like kohya ss. Maybe those guys make it right. Or something else.
@@NGIgri I did notice that they might have come up with a new version since this video was made. I wonder if there's a glitch in the new version. You can see that there's a couple differences. Like on my version you type in the resolution you want. But in this version in the video, you have to check off 512 or 1024.
@@brianmonarchcomedy yeah. And I found strange the resolution changes itself. I tried to train LoRA on 1024, but it automaticaly changes to 512. Maybe it's because of12gb of VRAM.
I have some errors trying to figure out but not sure if anybody had the same [INFO] RuntimeError: use_libuv was requested but PyTorch was build without libuv support [ERROR] Command exited with code 1.
for some reason doesn't see libuv did the same in other slower machine and works no idea what to do I try different version of pytorch and same.
I found the solution, basically is because I have two GPUs, so the script is trying to use both RTX3090 and for this reason, doesn't work.
I just added the extra calls " --num_processes=1 ^ --num_machines=1 ^" on the "train.bat" file and now is using only one GPU without any error.
I guess there must be something extra to be installed for multiple GPUs to work. Thanks again for the tutorials was very usefull
@@jocg9168 wow thanks for the info i also had 2 GPU's now it works
@@jocg9168 Having same issues. How do you add those extra calls to the train script? I edit and then it disappears?
Train.bat location? Not seeing that. Thanks for this info as it's definitely not with the pytorch etc. I'm a newb at this part.
How can we make seamless textures on Flux? I have trained my own lora but I can't get seamless results
are you using 1024 image ? ive tried this and comfyui flux training, the comfyui give me much better result (likeliness) than the flux gym, have you tried to compare lora from this and comfyui ? nice simple vid tho, gonna try it latter using your setting, and it seems like thee size of the dataset is also crucial ? (size as in bytes per image not in how many files)
Did you compare yet? I am about to do it too, but maybe you already got some experience?
@@ChrissyAiven yea tried it and this one had bad result i prefer ai toolkit Lora training using 1024px inage
@@TheLuc1890 ok thx but you can use 1024 px on comfy trainer too.
Finished after a couple of hours but didnt generate any actual lora output
I keep having the same thing happened. I trained maybe 7 models but only got my lora senor output model once
I have a 10Gb VRAM GPU (3080) I am getting out of memory error
Same here with 8GB
change your model to dev-fp8 ,that will be working
@@he-xs5le have you tried yourself?
@@lom1910 yes,I got the result
@@he-xs5le on 8gb right?
How to run it on already existing flux model, not fp16?
They are made for fp16 to train. As it mentioned. I haven't try gguf or other version.
@@TheFutureThinker I have 8GB VRAM and 16 RAM, but I instantly get out of memory with that default models. 😕 Pinokio either crashes completely or I get an error in the terminal
Not enough Vram , yes, it will not perform. And the trainer min. use 12gb vram.
Has the new kohya_ss update tutorial come so that it can work with the FLUX model?
I recently installed it, but even though it was the latest update, it did not support FLUX. If possible, teach us how to update it. thanks
@@mr.entezaee @6:00
See @5:50
@@Carl-md8pc good you pay attention 👍
not working
How are you getting them hands so good?
Just using Flux to generate it, haven't change much, you can see in the last part.
This is a lot simple webui thanks
Great video, can this run on a GeForce RTX 2060 super 8GB?
As long as 12, 16, 20GB vram , then its good to go. 8 gb you can try , but it might take you a day to run.
@@TheFutureThinker 😂 ok thanks.
@@TheFutureThinker when i try to run lora model training with 2060 6gb in 10 hours only 6% training complete only
with torch.enable_grad(), device_autocast_ctx, torch.cpu.amp.autocast(**ctx.cpu_autocast_kwargs): # type: ignore[attr-defined] What's the problem, please
Please try with GPU.
@@TheFutureThinker Thank you. What should I do to use the GPU, please
I love you!!!! thank you!!!!
Not working plus error😔
Kinda crazy how just vram adds 3x faster speed.
I just wish nvidia sold a 48g. Vram graphics card at a 4070ti price point. It costs them nothing to double vram. And given the price of 4070tj we got ripped off hard with 12gb vram. So sad. Even games in 4k you benefit from 24gb vram. They just want to upsell. Given the price of a 4090 it should be 48gb vram too
3080 can train flux now?
again, in other comment, don't care what model you have. 12, 16, 20 GB Vram that's what they have in the setting.
Can I interrupt it?
If you want to start over again. Then you can stop it in the middle of the train, but it won't resume.
Maybe show a few examples?
Great video.
Music suggestion: GuyJ
I don't artist music, its copyright.
i rather use stock, play by myself or generate
@@TheFutureThinker
Yes, that makes sense.
how to train sdxl lora?
search for the old video, I remember someone did it and I did it.
The character is a very simple example, I suppose transferring the artist’s style will require 200-300 initial scans, and the training time for 12-16 GB will increase to 72 hours?
I started a character lora on my 12Gb VRAM card. It was 25 images at 1024, 1500 steps and I think everything else was default. Once it got to the epoch part (which took awhile), the estimated time was 23hrs. I didn't continue but I will try again with less images, steps and epochs.
A real low GPU would be more like 4GB as many have old machines.
So we should define real low , AI low, server low
@@TheFutureThinker Roop almost runs on anything. They had fast SD CPU, but the trend went to high VRAM then now lower VRAM. There was the LCM LoRA models with Fast SD CPU. Most of this is just photoshop on steroids, but there should be more of a manual way of doing it locally for almost anyone that does not have the VRAM. IMO.
Thanks, I'll check with my 12Gb card, I think it will took not less than 5 hrs :)
you can try it. let us know how much it take for 12gb. :)
Ok, no, it's working, expected to finish in about 8 hrs :)
Right on😎👍 but....omg 8 hours... Put that on the cloud server or something.🥹
@@TheFutureThinker It tooks 8hrs 19 min
@@luisellagirasole7909 go to sleep take a rest, then it's done 😄
Beavis and Butthead dance during the metal music.😅
Hehe yeah yeah hehe yeah yeah... 😂😂
This thing doesn't work at all. All processing goes to my CPU rather than my GPU.
Your problem . My rx 7600 xt gpu is used to it's full potential and my cpu is just chilling.
@dariocardajoli6831 Okay, not sure how this contributed
@@jason54953 u didn't even say what's your GPU lol
@dariocardajoli6831 i have a rtx 4070 ti
12 Giga VRAM is still too high
I am trying with my 3090.
20gb 20 photo , 51mins in 4090 , let see what you have in 3090
@@TheFutureThinker I also have 4090 / how much data is enough for training with good results?
I got much better results using SimpleTrainer. And I think is more optimized than SD-script
Cool, i will try that one later thanks.
Low Vram.... 12GbVRAM 😂
(😢4GbVram😢)
same ... 3070 Ti 8Go😢😭
Dude just give 0.80 usd to replicate and be done with your loras fast
Yes, good solution for entry level.
@@TheFutureThinker tried a couple, no fine tuning of course. It was alright for charactera but did not try entire styles, quess the key words, having the text files and training it with more detail might have better results.
Once I hit training, at the bottom I'm seeing :
[INFO] RuntimeError: use_libuv was requested but PyTorch was build without libuv support
[ERROR] Command exited with code 1
[INFO] Runner:
Out of my element, like Donny, on this one. Do I have to install it with this?
pip install torch --no-cache-dir
If anyone has some insight. Would appreciate it. Everything works perfect until this point.