From 7caf9441812dd8db9884e9b5b8205f3fc3d23d2c Mon Sep 17 00:00:00 2001 From: Daniel Bevenius Date: Mon, 4 Nov 2024 07:11:53 +0100 Subject: [PATCH] fix flake8 errors --- convert_hf_to_gguf.py | 6 ++++-- 1 file changed, 4 insertions(+), 2 deletions(-) diff --git a/convert_hf_to_gguf.py b/convert_hf_to_gguf.py index 8b89afbd4c92c..21f861917e5e9 100755 --- a/convert_hf_to_gguf.py +++ b/convert_hf_to_gguf.py @@ -1662,6 +1662,7 @@ def prepare_tensors(self): if len(experts) > 0: raise ValueError(f"Unprocessed experts: {experts}") + @Model.register("MllamaForConditionalGeneration") class MLlamaModel(Model): model_arch = gguf.MODEL_ARCH.MLLAMA @@ -1673,10 +1674,11 @@ def find_hparam(self, keys: Iterable[str], optional: bool = False) -> Any: return super().find_hparam(keys, optional) def set_vocab(self): - logger.info(f"[danbev] set_vocab...") + logger.info("[danbev] set_vocab...") def set_gguf_parameters(self): - logger.info(f"[danbev] set_gguf_parameters...") + logger.info("[danbev] set_gguf_parameters...") + @Model.register("BitnetForCausalLM") class BitnetModel(Model):