Replies: 2 comments 3 replies
-
Hi, SD released two types of weights, 4.3 Gb and 7.7Gb (the larger weights have ema checkpoints included). I am using the 4.3 Gb weights. I am guessing that the 2Gb weights recently announced are just the 4.3 Gb weights converted to half precision which effectively reduces their size from 4.3Gb --> 2.1Gb. But I am not sure since they may be doing other optimizations too. I am already converting the weights to half-precision in my code, so if that is the case, I don't think my code will run faster due to the reduction in weight size. To reduce the required VRAM further, this repo uses other modifications like splitting the model into multiple parts and moving them into the GPU only when required. This increases the inference time a little bit, but one can generate larger images. |
Beta Was this translation helpful? Give feedback.
-
Hi Basu, Thank you for your kind reply, this is very interesting information and much appreciated. I am also a fellow RTX 2060 owner, and am extremely interested in using your SD optimization (I've been a happy beta tester and fell in love with the technology). I have a few extra questions, if I may:
Thanks again, and keep up the great work! Exciting times, for sure! |
Beta Was this translation helpful? Give feedback.
-
Hi,
I can't wait to try your code when the weights are finally released! However, I'm curious to know which weights you're using, as it seems that the ones that were sent to testers are bigger (7.1 gb) than the ones that will be released within a couple of days. SD staff announced the final optimized weights would be around 2 gb in size. Does that mean that your code could run even faster on such a small weight? Or is it already the one you make your tests with?
Cheers!
Beta Was this translation helpful? Give feedback.
All reactions