Performance optimization of SDXL inference pipeline #1104
bssrdf
started this conversation in
Show and tell
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Over the past a couple of months, I have been trying to speed up SDXL inference, motivated by #772. Here is a summary of what I have achieved so far.
First, this is what a recent master (commit 8f6c5c2) can do

My optimized pipeline

Combining roughly these 5 optimizations above, on 4090, I can do
12.5it/s19it/s8.5it/s11.59it/s6.35it/s 6.5it/s 7.17it/s 7.6it/s7.99it/sAs a side project, I also tried to improve FLUX inference pipeline.
@JustMaier, not sure if you and your team are still interested in pursuing using sd.cpp. If so, please let me know and you can give it a try in your production environment. Thanks.
Beta Was this translation helpful? Give feedback.
All reactions