Speed and optimisation #24
-
I'm starting this as Q&A because its mainly a question I've been wondering about: A friend of mine for example is doing this on a GTX 960 (what a madman) and he's experiencing up to 3 times the speed when doing inference in ComfyUI over Automatic's. And then in addition, complicated workflows with more advanced conditionings get even slower (which is to be expected). |
Beta Was this translation helpful? Give feedback.
Replies: 3 comments 2 replies
-
Try using an fp16 model config in the CheckpointLoader node. That should speed things up a bit on newer cards. It should be at least as fast as the a1111 ui if you do that. If it isn't let me know because it's something I need to fix. |
Beta Was this translation helpful? Give feedback.
-
Found this fix for Automatic1111 and it works for ComfyUI as well. I have RTX 4070Ti a had this gains with update to cudnn8800 (Model ChilloutMix, 512x512, 20 steps, CFG 8, Euler): |
Beta Was this translation helpful? Give feedback.
-
I think this can be closed because you can just use the CheckpointLoaderSimple now to get optimal performance. |
Beta Was this translation helpful? Give feedback.
I think this can be closed because you can just use the CheckpointLoaderSimple now to get optimal performance.