torch.compile with stable diffusion #7828
Unanswered
Madhumitha-MCW
asked this question in
Q&A
Replies: 4 comments 2 replies
-
|
Beta Was this translation helpful? Give feedback.
0 replies
-
Yes, the time taken to compile vae is more than only compiling unet part. Thanks on your reply! |
Beta Was this translation helpful? Give feedback.
1 reply
-
Yes, I just was mentioning the first inference time(compilation part). The final inference is not slow. |
Beta Was this translation helpful? Give feedback.
1 reply
-
Yes, it would. I first thought the entire inference time was longer, but when I verified it, it was just the initial part which took longer |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I would like to know why any inference example with stable diffusion has only unet part of the model passed into torch.compile(). https://pytorch.org/TensorRT/tutorials/_rendered_examples/dynamo/torch_compile_stable_diffusion.html
When using torch.compile with model.vae it relatively takes longer time than just compiling unet part. So, is it recommended to compile only unet or what is the exact reason?
And if I want to execute in graph mode, will all other uncompiled parts undergo eager execution if I compile only unet part?
Beta Was this translation helpful? Give feedback.
All reactions