Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

3221225477, can't load checkpoints sucessed. #1

Open
wikeeyang opened this issue Nov 20, 2024 · 14 comments
Open

3221225477, can't load checkpoints sucessed. #1

wikeeyang opened this issue Nov 20, 2024 · 14 comments

Comments

@wikeeyang
Copy link

wikeeyang commented Nov 20, 2024

test load flux.1 dev bf16 and fp8 safetensors file with your "universal_modulator.safetensors" together in "FluxModCheckpointLoader" node, can't load sucessed. error message is:3221225477, webui lose connect with server, need to reatsrt ComfyUI server.
15580efb73420ed03dc92be0cddecde

@lodestone-rock
Copy link
Owner

use base flux dev / schnell instead of quantized version
https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/flux1-dev.safetensors
https://huggingface.co/black-forest-labs/FLUX.1-schnell/blob/main/flux1-schnell.safetensors

currently the codebase does not support loading the fp8 weight directly
quant mode option will quantize the model on the fly

@lodestone-rock
Copy link
Owner

also can you share the entire error log?

@wikeeyang
Copy link
Author

Yes, I try the https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/flux1-dev.safetensors model file, the same error.
1732089076401

@wikeeyang
Copy link
Author

I know the node should load the model, merge and convert it in memory, but may be the dev model had loaded, the error has occurred before merge and convert. are there any requirements for a GPU? For example, BF16 must be supported?

@wikeeyang
Copy link
Author

wikeeyang commented Nov 20, 2024

can you try to export 16bit or 8bit unet model?
1732089850398

@Bananapuncakes
Copy link
Collaborator

Bananapuncakes commented Nov 20, 2024

It seems like you're trying to create a checkpoint which isn't going to work. The model has a completely different architecture hence the requirement for this custom node in the first place. Is there a particular reason for as to why you're trying to save the checkpoint here?

@wikeeyang
Copy link
Author

No, Just for test, due to run your workflow error and stop on "FluxModCheckpointLoader" node.
I have 64GB RAM, 24GB GPU. Windows 11 x64, Python 3.11 torch 2.5.1 and CUDA 12.4.

@wikeeyang
Copy link
Author

and I can load offical Flux1-dev.safetensors convert online to fp8 use below node:
1732091104700

@wikeeyang
Copy link
Author

My env is no problem. tested as below:
1732091368127

@Bananapuncakes
Copy link
Collaborator

Bananapuncakes commented Nov 20, 2024

What GPU are you currently using and how much VRAM does it have? I don't see anything out of the ordinary in the logs that you provided earlier. We're going to need more detailed information than what you gave.

@wikeeyang
Copy link
Author

Nvidia P40, 24GB VRAM. but as i looked, the node load to CPU first, may be don't start to convert and transfer to GPU, the error occurred.

@wikeeyang
Copy link
Author

wikeeyang commented Nov 20, 2024

may be here somthing need to change?
1732092872024
1732092983611

@Bananapuncakes
Copy link
Collaborator

Bananapuncakes commented Nov 20, 2024

Ah yes, that would be it since currently the model is only setup for bf16 at the moment which requires a higher compute capability (I think 8.0?) whilst the P40 would only support fp16.
I'll talk to lodestone tomorrow about potentially adding fp16 support.
@lodestone-rock

I'm not sure if ComfyUI is doing an on the fly conversion for Flux since the model is bf16 on Hugging Face. I'm guessing default may be either bf16 or fp16.
CleanShot 2024-11-20 at 20 09 37@2x

@wikeeyang
Copy link
Author

Yes, ComfyUI can do the fly conversion from bf16 to fp16 or fp8e4m3......

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants