TensorRT & Flux Dev #4484
-
@comfyanonymous What am I doing wrong? TensorRT doesn't support 16gb VRAM? Thank you in advance |
Beta Was this translation helpful? Give feedback.
Replies: 8 comments 8 replies
-
can someone please help? |
Beta Was this translation helpful? Give feedback.
-
TensorRT needs more than 24GB vram at the moment to convert a Flux model, even a 4090 isn't enough. |
Beta Was this translation helpful? Give feedback.
-
couldn't split it to chunks and save it , in loop for gpu and cpu |
Beta Was this translation helpful? Give feedback.
-
I am going to get so mad if Nvidia doesn't start putting out affordable 48Gb cards soon come on even a 4090 ti with 36gb vram 20.000 tensors would be great. |
Beta Was this translation helpful? Give feedback.
-
Couldn't someone rent a H100 to convert it ? |
Beta Was this translation helpful? Give feedback.
-
I could convert it at my workplace but if I remember correctly the TRT will be specific to the GPU I used and not portable to another? I tried it with the schnell fp8 checkpoint and I run into an error: [09/24/2024-07:07:26] [TRT] [E] IBuilder::buildSerializedNetwork: Error Code 9: API Usage Error (Networks with BF16 precision require hardwar e with BF16 support.) |
Beta Was this translation helpful? Give feedback.
-
You can try converting the Flux model using a ada 6000 and then run the engine with a 4090. However this only works if nothing else is running on the 4090. This means no monitor plugged into it and no applications running on it. Alternatively you can try fp8 + --fast + the torch compile node. |
Beta Was this translation helpful? Give feedback.
-
I compiled flux with tensorrt, using the defaults in the tensorrt node pack, and observed better performance*.
On RTX A5000, pytorch 2.4.1, tensorrt 10.5.0 Windows |
Beta Was this translation helpful? Give feedback.
TensorRT needs more than 24GB vram at the moment to convert a Flux model, even a 4090 isn't enough.