- Notifications
You must be signed in to change notification settings - Fork3k
[Precomputed ControlNet] Speed up ControlNet by 45% - but is it necessary?#216
Uh oh!
There was an error while loading.Please reload this page.
Uh oh!
There was an error while loading.Please reload this page.
-
Hi everyone, we plan to experiment with a feature called "Precomputed ControlNet". This can be achieved by modifying 2 or 3 lines of the training code to progressively disconnect the input concat here: By doing this, we will be able to execute controlnet only one time before diffusion, rather than in every diffusion iteration. The controlnet should beequally powerful and robust as before. This will lead to a speed up by about 40% to 45%. And this will further decrease the required GPU memory. Nevertheless, if we can observe any performance decrease (even minimal decrease) in any experimental setting (including no prompt setting, short prompt setting, and long prompt setting), we will make this experimental feature on-hold to avoid confusing and to prevent giving new users a mis-estimate of the models' capabilities. Let us know what you think about it! Thank you for your support as always. Update (after all experiments of cn1.1): this experiment fails to train ControlNets as good as the proposed implementation in our paper. We observe that models trained with this method tends to produce more artifacts and less robust results. We have given up this feature. The input from each diffusion steps is necessary for robustness, and necessary for special models like Shuffle and IP2P (in controlnet 1.1). |
BetaWas this translation helpful?Give feedback.
All reactions
👍 60😄 4❤️ 17🚀 3
Replies: 19 comments 1 reply
-
Great work so far. This has been a lot of fun to play with. |
BetaWas this translation helpful?Give feedback.
All reactions
👍 1
-
I am down to try it out |
BetaWas this translation helpful?Give feedback.
All reactions
-
Good idea, afaik many ppl w/ low vram GPUs are struggling w/ using CN w/ highres fix so if it works it'll be great. |
BetaWas this translation helpful?Give feedback.
All reactions
-
Great which 3 lines did you change? I would like to try it. |
BetaWas this translation helpful?Give feedback.
All reactions
-
looks awesome , keep it up |
BetaWas this translation helpful?Give feedback.
All reactions
-
Is this speedup competitive with T2I adapter? |
BetaWas this translation helpful?Give feedback.
All reactions
-
that's good news for me, 期待早点体验,目前gpu vram总是不够,计算也很慢 |
BetaWas this translation helpful?Give feedback.
All reactions
-
Wouldn't this let you use a much larger model |
BetaWas this translation helpful?Give feedback.
All reactions
-
autoreply:Your mail has been received, and I will reply as soon as I canBest Regards;) |
BetaWas this translation helpful?Give feedback.
All reactions
-
Yes, it's really necessary. I have a 4GB GPU and I would benefit of this. Thank you for you hard work. |
BetaWas this translation helpful?Give feedback.
All reactions
-
Good idea, I need it very much, |
BetaWas this translation helpful?Give feedback.
All reactions
Uh oh!
There was an error while loading.Please reload this page.
Uh oh!
There was an error while loading.Please reload this page.
-
Could we use the released model to achieve this or should we retrain a new model to achieve this? |
BetaWas this translation helpful?Give feedback.
All reactions
-
we need it =)) |
BetaWas this translation helpful?Give feedback.
All reactions
-
how to deal with the timesteps? Does it use T at the first step, and the T result for all following steps? |
BetaWas this translation helpful?Give feedback.
All reactions
👀 1
-
autoreply:Your mail has been received, and I will reply as soon as I canBest Regards;) |
BetaWas this translation helpful?Give feedback.
All reactions
-
Update: this experiment fails to train ControlNets as good as the standard implementation. We have given up this feature. The input from each diffusion steps is necessary for robustness, and necessary for special models like Shuffle and IP2P. |
BetaWas this translation helpful?Give feedback.
All reactions
👀 2
-
can we see the images? |
BetaWas this translation helpful?Give feedback.
All reactions
-
autoreply:Your mail has been received, and I will reply as soon as I canBest Regards;) |
BetaWas this translation helpful?Give feedback.
All reactions
Uh oh!
There was an error while loading.Please reload this page.
Uh oh!
There was an error while loading.Please reload this page.
-
can we use NAS (neural architecture search) for even more GPU efficiency ? finding new neural architecture which is even much faster ? does anyone have Nvidia dgx workstation to run NAS , and find a better and even more gpu efficient and fast algorithm ? |
BetaWas this translation helpful?Give feedback.
All reactions
❤️ 2
-
autoreply:Your mail has been received, and I will reply as soon as I canBest Regards;) |
BetaWas this translation helpful?Give feedback.
