Movatterモバイル変換


[0]ホーム

URL:


Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Sign up
Appearance settings

[Precomputed ControlNet] Speed up ControlNet by 45% - but is it necessary?#216

Discussion options

Hi everyone, we plan to experiment with a feature called "Precomputed ControlNet". This can be achieved by modifying 2 or 3 lines of the training code to progressively disconnect the input concat here:

image

By doing this, we will be able to execute controlnet only one time before diffusion, rather than in every diffusion iteration. The controlnet should beequally powerful and robust as before. This will lead to a speed up by about 40% to 45%. And this will further decrease the required GPU memory.

Nevertheless, if we can observe any performance decrease (even minimal decrease) in any experimental setting (including no prompt setting, short prompt setting, and long prompt setting), we will make this experimental feature on-hold to avoid confusing and to prevent giving new users a mis-estimate of the models' capabilities.

Let us know what you think about it! Thank you for your support as always.

Update (after all experiments of cn1.1): this experiment fails to train ControlNets as good as the proposed implementation in our paper. We observe that models trained with this method tends to produce more artifacts and less robust results. We have given up this feature. The input from each diffusion steps is necessary for robustness, and necessary for special models like Shuffle and IP2P (in controlnet 1.1).

You must be logged in to vote

Replies: 19 comments 1 reply

Comment options

Great work so far. This has been a lot of fun to play with.

You must be logged in to vote
0 replies
Comment options

I am down to try it out

You must be logged in to vote
0 replies
Comment options

Good idea, afaik many ppl w/ low vram GPUs are struggling w/ using CN w/ highres fix so if it works it'll be great.

You must be logged in to vote
0 replies
Comment options

Great which 3 lines did you change? I would like to try it.

You must be logged in to vote
0 replies
Comment options

looks awesome , keep it up

You must be logged in to vote
0 replies
Comment options

Is this speedup competitive with T2I adapter?

You must be logged in to vote
0 replies
Comment options

that's good news for me, 期待早点体验,目前gpu vram总是不够,计算也很慢

You must be logged in to vote
0 replies
Comment options

Wouldn't this let you use a much larger model

You must be logged in to vote
0 replies
Comment options

autoreply:Your mail has been received, and I will reply as soon as I canBest Regards;) 
You must be logged in to vote
0 replies
Comment options

Yes, it's really necessary. I have a 4GB GPU and I would benefit of this. Thank you for you hard work.

You must be logged in to vote
0 replies
Comment options

Good idea, I need it very much,

You must be logged in to vote
0 replies
Comment options

Could we use the released model to achieve this or should we retrain a new model to achieve this?
I tried to tweak the sampling steps in the way you described and used the released model, and it didn't work.😢

You must be logged in to vote
0 replies
Comment options

we need it =))

You must be logged in to vote
0 replies
Comment options

how to deal with the timesteps? Does it use T at the first step, and the T result for all following steps?

You must be logged in to vote
0 replies
Comment options

autoreply:Your mail has been received, and I will reply as soon as I canBest Regards;) 
You must be logged in to vote
0 replies
Comment options

lllyasviel
May 6, 2023
Maintainer Author

Update: this experiment fails to train ControlNets as good as the standard implementation. We have given up this feature. The input from each diffusion steps is necessary for robustness, and necessary for special models like Shuffle and IP2P.

You must be logged in to vote
1 reply
@Njasa2k
Comment options

can we see the images?

Comment options

autoreply:Your mail has been received, and I will reply as soon as I canBest Regards;) 
You must be logged in to vote
0 replies
Comment options

can we use NAS (neural architecture search) for even more GPU efficiency ?

finding new neural architecture which is even much faster ?

does anyone have Nvidia dgx workstation to run NAS , and find a better and even more gpu efficient and fast algorithm ?

You must be logged in to vote
0 replies
Comment options

autoreply:Your mail has been received, and I will reply as soon as I canBest Regards;) 
You must be logged in to vote
0 replies
Sign up for freeto join this conversation on GitHub. Already have an account?Sign in to comment
Labels
None yet
15 participants
@lllyasviel@axsddlr@lioo717@OedoSoldier@yakuzadave@drbobo0@TheLukaDragar@neverix@Georgefwt@karkra911@dan4ik94@ffdown@Aridea2021@Njasa2k@xueqing0622

[8]ページ先頭

©2009-2025 Movatter.jp