I listed all diffusion model, vae link in this blog post, so how YT don't like HF links in description. thefuturethinker.org/nvidia-sana-in-comfyui-setup-tutorial-guide/
Please provide us working vae link. All people having problem with grey,black,blurry colorful images. Someone told that oldest version working, give it a chance and inform us.
Don’t even bother watching videos from a month ago 😂, honestly every time I open my TH-cam a new major model is out including 2x video models. New apps, new quants, new great workflows, new control nets 😂 I can’t download fast enough
If you get black you can get the oldest vae from the commit history on hugging face. There is only one vae file in that commit and it works at the moment.
@@Pernicuz I downloaded what I thought would have been the oldest version, but without success, still a black image. could you guys please leave a direct link here ? thanks a lot
Thanks for the pointer it worked, this should be the top comment. Download the earliest vae model and it works, looks like they broke something in the meantime.
i don't think that this is the VAE, the Ksampler shows no preview at all and all samples i have seen have ksampler preview. The extra models are simply broken.
I did everything as in the video. I downloaded all the databases. It is generated without errors , but the end result is a black square (( ComfyUI has been updated. Does not help. ( config win 10/10400/64 gb/ 3080 10gb (latest drivers)
The error "Expected all tensors to be on the same device, but found at least two devices, cpu and cuda:0! (when checking argument for argument mat1 in method wrapper_CUDA_addmm)" appears. How can I fix this? For your information, I am using an RTX 3080.
Nerdy Rodent was required to remove his Hunyuan video since he is in the UK and there are EU and UK license issues with that model. This is video too ?
I am only getting black or grey images. This error keeps appearing in console: Unused kwargs: ['_load_in_4bit', '_load_in_8bit', 'quant_method']. These kwargs are not used in .
I wonder how inpainting and attention to details works with Sana. As far as I’ve seen in this video, Sana mostly fails to correctly represent objects and ignores generation of small details, like skin texture (woman’s example). But if more high quality FLUX-like results will be introduced, it’ll be amazing to have such a fast model to play with
@@TheFutureThinker This is more exciting than a new car, bro. Cause it's freedom to create. That would include a car design with no limits. I want to make smoke paint. It's possible with AI. Thanks for the vids.
hi, i don't know what hapens, always displays error "No package metadata was found for bitsandbytes" , i installed bitsandbytes and reinstalled torch and cuda and the error stays there
I got error GemmaLoader Can't use dtype 'torch.float16' with CPU! Set dtype to 'default'. I have set it to default, but i got error "The checkpoint you are trying to load has model type `gemma2` but Transformers does not recognize this architecture. This could be because of an issue with the checkpoint, or because your version of Transformers is out of date." Can you help me?
Hello, Im new to all this, and Im getting a bit of an error after following your steps, this is the error and I have no idea how to fix it, Input type (torch.cuda.HalfTensor) and weight type (torch.HalfTensor) should be the same, might you have an idea how to fix this and help a newbie out. any help is greatly appreciated... so it seems its telling me my model is running on GPU and data is on CPU, not sure how to make the adjustment to have both the same, any ideas, or is there a GPU work flow for this
Great video! thank you. I'm getting blank images generated even though everything is set up properly. Any ideas why? Edit: Ijust noticed everyone else has already reported it.. oh! and Nvidia's license is extremely non user-friendly...shame
Nice video, but please, when you do this kind of videos, explain how you do things. I´m searching for a way to search inside ComfyUI and I don´t find anyway to do it, I´m stuck almost at the beginning. Edit: Ok, I found a way and now I don´t get how to do the rest because you don´t explain how to do it, so it´s not a comprehensive tutorial as you call it, it´s a tutorial for people that knows what to do.
As its about 4 times smaller database than flux it pretty much cant be better and as variable. What comes to VRAM they should publish specific checkpoints. Like people, cats, dogs, buildings, landscapes etc.. 4 sure its faster. So is SDXL. Better in quality? Nope. More variance? Nope. I hate the hype going around everything released just to get clicks.
i am getting two errors back to back "Expected all tensors to be on the same device, but found at least two devices, cpu and cuda:0! (when checking argument for argument mat1 in method wrapper_CUDA_addmm" , "Expected all tensors to be on the same device, but found at least two devices, cpu and cuda:0! (when checking argument for argument mat1 in method wrapper_CUDA_addmm
Followed exactly but get "`.to` is not supported for `4-bit` or `8-bit` bitsandbytes models. Please use the model as it is, since the model has already been set to the correct devices and casted to the correct `dtype`."
Make sure got the right VAE for the model. I can't post the HF links on the description, so how YT don't like it. And I list it all in my blog post. thefuturethinker.org/nvidia-sana-in-comfyui-setup-tutorial-guide/
The downside with Comfyui is not primarily the cluttered UI, but the difficulty in gauging the settings for an optimal output. Trial and error takes precious time, and there are no standardized and proved settings for the best results. Opinions are like rear ends.
But I will let SANA sit for a month while you guys work out the kinks. Too much too fast. BTW 16GB cards will be the minimum for 2025. Buy a new GPU before those orange man tariffs kick in.
I listed all diffusion model, vae link in this blog post, so how YT don't like HF links in description.
thefuturethinker.org/nvidia-sana-in-comfyui-setup-tutorial-guide/
Please provide us working vae link. All people having problem with grey,black,blurry colorful images. Someone told that oldest version working, give it a chance and inform us.
The fastest at making a blank image! RIP ExtraVAELoader.
Damn, the pace of AI development right now is just ridiculous. Something I learned 3-6 months ago is already outdated.
@@dhanangright
@dhanang sometimes wake up another, a new thing release
3-6 Months?? I'd say 3-6 Weeks max 😅
@@maknien3-6 days😂
Don’t even bother watching videos from a month ago 😂, honestly every time I open my TH-cam a new major model is out including 2x video models. New apps, new quants, new great workflows, new control nets 😂 I can’t download fast enough
If you get black you can get the oldest vae from the commit history on hugging face. There is only one vae file in that commit and it works at the moment.
yup that worked for me too, thanks!
@@Pernicuz I downloaded what I thought would have been the oldest version, but without success, still a black image. could you guys please leave a direct link here ? thanks a lot
Thanks for the pointer it worked, this should be the top comment.
Download the earliest vae model and it works, looks like they broke something in the meantime.
@@kaiserscharrman Tried leaving a link but YT immediately deletes the comment.
The commit number is 38ebe9b227c30cf6b35f2b7871375e9a28c0ccce
i don't think that this is the VAE, the Ksampler shows no preview at all and all samples i have seen have ksampler preview. The extra models are simply broken.
i LOVE that your hobby came through in this video!!!!
I am not just tech and AI 😄😅
Great video! Thanks! 🙏
I did everything as in the video. I downloaded all the databases. It is generated without errors , but the end result is a black square ((
ComfyUI has been updated. Does not help. (
config win 10/10400/64 gb/ 3080 10gb (latest drivers)
I'm also only getting black images
Same. I just get grey square.
I got a black gray image too... not sure what I missed... did exactly as the video
same here
At least I don't feel like the only one hahaha
I'm just getting a pixelated mess. 😥
The error "Expected all tensors to be on the same device, but found at least two devices, cpu and cuda:0! (when checking argument for argument mat1 in method wrapper_CUDA_addmm)" appears. How can I fix this?
For your information, I am using an RTX 3080.
@@moviecartoonworld4459 I get the same error. Couldn't find a solution.
I was exposed to the same message with the general video card in my device RTX3050 4gb
How did you solve the problem, please?
Competitive Aquarium Design?!!!😱 I didn't know that was a thing!!!.... Oh... Sana is pretty cool too 😁👍
@UnclePapi_2024 search it, IAPLC 😉 you might be addicted to this hobby if you like nature, water, animals, and plants.
i get a lot of Missing VAE keys errors, and only black or grey generation (ex. ['encoder.project_in.weight',......)
Same here - and look at a thread a bit above from this - many users are getting this missing VAE keys message
same
@@parthwagh3607 A eff ton of errors then the ksampler says AttributeError: module 'xformers.ops.fmha' has no attribute 'BlockDiagonalMask'
Than you for taking the time to explain it !
Censored and sanitized for my protection?
Without finishing the video, how are the hands and realism in general?
just black image
Nerdy Rodent was required to remove his Hunyuan video since he is in the UK and there are EU and UK license issues with that model.
This is video too ?
Are generated images allowed for commercial use? From what I read in the license files, they are not permitted.
NO, the license clearly says "non commercial use only". so I don't even bother with the black rectangle it generates due to an VAE issue any longer...
I am only getting black or grey images.
This error keeps appearing in console:
Unused kwargs: ['_load_in_4bit', '_load_in_8bit', 'quant_method']. These kwargs are not used in .
I wonder how inpainting and attention to details works with Sana. As far as I’ve seen in this video, Sana mostly fails to correctly represent objects and ignores generation of small details, like skin texture (woman’s example). But if more high quality FLUX-like results will be introduced, it’ll be amazing to have such a fast model to play with
Well, its a base model and very small size model. So it happened, some type of image not showing detail.
@ hopefully the infrastructure to fine-tune Sana and use it with other generation pipelines components will be developed soon:))
I was literally looking at NVDA for a long lol. This is awesome.
Haha 😂😂
But this model still in early stage. Wait for the fine tune.
@@TheFutureThinker This is more exciting than a new car, bro. Cause it's freedom to create. That would include a car design with no limits. I want to make smoke paint. It's possible with AI. Thanks for the vids.
hi, i don't know what hapens, always displays error "No package metadata was found for bitsandbytes" , i installed bitsandbytes and reinstalled torch and cuda and the error stays there
Are you using Windows?
I have the same error
with windows 11 4060 16gb ti
I got error GemmaLoader Can't use dtype 'torch.float16' with CPU! Set dtype to 'default'. I have set it to default, but i got error "The checkpoint you are trying to load has model type `gemma2` but Transformers does not recognize this architecture. This could be because of an issue with the checkpoint, or because your version of Transformers is out of date." Can you help me?
The next step is when some company launch a common AI model. For example ... use resource from sd 1.5 in Pony, SDXL, or other model.
Thank you. Hopefuly, we will be moving from flux to other models.
what about sana vs flux dev? can sana win? any do tests alredy?
Is there a way to have an output of various cfg levels? I'd think it makes sense to have an array of the different variants.
like others same issue on windows 11: GemmaLoader, No package metadata was found for bitsandbytes
Thank You for this video! :)
Hope you like it 😄
Hello, Im new to all this, and Im getting a bit of an error after following your steps, this is the error and I have no idea how to fix it, Input type (torch.cuda.HalfTensor) and weight type (torch.HalfTensor) should be the same, might you have an idea how to fix this and help a newbie out. any help is greatly appreciated... so it seems its telling me my model is running on GPU and data is on CPU, not sure how to make the adjustment to have both the same, any ideas, or is there a GPU work flow for this
Thanks, great ideo. shame these models probably don't have controlnets yet?
Is there any news on SANA Lora training?
Does this work with Mac considering it’s Nvidea?
yeah, thanks for the updates!!
Try this out😉
Great video! thank you. I'm getting blank images generated even though everything is set up properly. Any ideas why? Edit: Ijust noticed everyone else has already reported it.. oh! and Nvidia's license is extremely non user-friendly...shame
How is it with generating text in images ?
I keep getting this at the ksampler: KSampler 'int' object is not subscriptable
I got the same error when I used the whole number 1 .... 1.1 works, so it must be some scripting error
Maybe a TensorRT version next ?
what's left is to make them read our thoughts xD Typing cost to much time haha
Nice video, but please, when you do this kind of videos, explain how you do things. I´m searching for a way to search inside ComfyUI and I don´t find anyway to do it, I´m stuck almost at the beginning. Edit: Ok, I found a way and now I don´t get how to do the rest because you don´t explain how to do it, so it´s not a comprehensive tutorial as you call it, it´s a tutorial for people that knows what to do.
Nsfw? I might try this for up-res of flux fills
Haha, i remember our office with a 180cm planted tank 😂 when will you setup a new layout again?
Maybe coming 2025 I join again the IAPLC 😁🤫. And I need to buy some new stone. Too bad AI cannot generate that. LOL
Can it do hands well? Also I guess loras will come to civitai soon enough.
yes, im currently developing loras for human realism, the flux process seems hundreds of times longer than this new one from NVIDIA, im happy.
@@masterwillian7785nice 👍 please keep us update on your Lora for Sana.
@@masterwillian7785 That sounds great. Will watch your video on the lora when you release it.
Mindblowing o_O
Excellent explained like everytime! ❤
Thanks
The question every user wants to ask, and no developer wants to answer:
"Does it do hands?"
Thanks it works
output were grey photos even followed your instructions in install on Window
Hope it supports lora, controlnet and ipadapter soon
As its about 4 times smaller database than flux it pretty much cant be better and as variable. What comes to VRAM they should publish specific checkpoints. Like people, cats, dogs, buildings, landscapes etc.. 4 sure its faster. So is SDXL.
Better in quality? Nope. More variance? Nope.
I hate the hype going around everything released just to get clicks.
Do LORAs work with this?
brooo waiting on aqua scaping channel now :D
I know in your place that have beautiful wood 😁 and good water condition for fish
It is wonderful for creating landscapes but for creating humans it is very bad, the worst.
Now just need a controlnet for it
CPU wahh🤩
@@golddiggerprankz yup yup the text encoder do the same as T5 but very lightweight.
not for generating the image, only for the prompt.
nice tutorial. but it only generates a solid gray image.
Did this work for anyone?
i am getting two errors back to back "Expected all tensors to be on the same device, but found at least two devices, cpu and cuda:0! (when checking argument for argument mat1 in method wrapper_CUDA_addmm" , "Expected all tensors to be on the same device, but found at least two devices, cpu and cuda:0! (when checking argument for argument mat1 in method wrapper_CUDA_addmm
After 3 hours of googling and re-installing - yes. But... This model feels like proof of concept, not real thing.
I have try Sana and i prefer by far flux, something is just not right in the lighting and details.
@@petertremblay3725 same with redux and pulid flux still rules
Followed exactly but get "`.to` is not supported for `4-bit` or `8-bit` bitsandbytes models. Please use the model as it is, since the model has already been set to the correct devices and casted to the correct `dtype`."
updated comfyui today and it's fixed but now I get the black screen. Will try to find that old vae.
Yea so you lost me at search box when you moved the mouse around.
i got only a solid grey colored picture ...
Same here. Everything is updated, but no image.
Make sure got the right VAE for the model. I can't post the HF links on the description, so how YT don't like it.
And I list it all in my blog post. thefuturethinker.org/nvidia-sana-in-comfyui-setup-tutorial-guide/
getting same result did anyone fix it yet ? btw i had downloaded the same vae from the link same as the model
@@TheFutureThinker seems several are having the same issue. Still happening after verifying VAE
I dont know if it is related but google/gemma-2-2b-it does not have access to install.
iit works but not sure it bether than sd15 or sdxl :(
Ai Never sleeps, dam!
😂😂😂yup
The downside with Comfyui is not primarily the cluttered UI, but the difficulty in gauging the settings for an optimal output. Trial and error takes precious time, and there are no standardized and proved settings for the best results. Opinions are like rear ends.
How about people?
6:55
But I will let SANA sit for a month while you guys work out the kinks. Too much too fast. BTW 16GB cards will be the minimum for 2025. Buy a new GPU before those orange man tariffs kick in.
Censored.
I
All I get is a blnk picture on every run.