Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

how about make load diffusion model+ support gguf #27

Open
xueqing0622 opened this issue Jan 10, 2025 · 4 comments
Open

how about make load diffusion model+ support gguf #27

xueqing0622 opened this issue Jan 10, 2025 · 4 comments

Comments

@xueqing0622
Copy link

how about make load diffusion model+ support gguf
there are example load model include gguf,
https://github.com/blepping/ComfyUI_FluxMod/tree/feat_gguf
lodestone-rock/ComfyUI_FluxMod#15

@blepping
Copy link
Contributor

I'm not sure it would make sense. The difference between the loaders is being able to set some extra dtypes. The advanced GGUF loader lets you enter the dequant and patch dtypes:

image

You can't enter something like FP8 there, but I'm not sure how that would work. Dequant to fast FP8 might be possible, but dequantizing to another quantized format is probably going to be a big quality loss (and likely would need support implemented in ComfyUI-GGUF). It also would only work on GPUs that support the FP8 ops with ComfyUI's --fast parameter.

@chengzeyi
Copy link
Owner

Do you mean converting gguf models to native format so that it could work with ComfyUI better?

@xueqing0622
Copy link
Author

xueqing0622 commented Jan 10, 2025

I'm using gguf which works just like normal safetensors, and it seems that the threshold has a lot to do with the number of steps,
#19 ,
this morning I've heard from videos that the load diffusion model+ has better support for lora. So I asked if load diffusion model+ is ready to support gguf.
image

@chengzeyi
Copy link
Owner

I'm using gguf which works just like normal safetensors, and it seems that the threshold has a lot to do with the number of steps,

#19 ,

this morning I've heard from videos that the load diffusion model+ has better support for lora. So I asked if load diffusion model+ is ready to support gguf.

image

yes support for Lora should be better since it fuses the Lora into the weights which makes the inference faster

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants