Support for GGML Files of Falcon 40b? #418
Replies: 3 comments 2 replies
-
ctransformers just added support for this and it seems to be working. It's been my goto for any model that doesn't work in llama https://github.com/marella/ctransformers/tree/main I've got a space setup using it based on the Falcon demo here: https://huggingface.co/spaces/matthoffner/falcon-mini 40b version: https://huggingface.co/spaces/matthoffner/falcon-fastapi |
Beta Was this translation helpful? Give feedback.
-
Thanks a lot Matt. This is very helpful.
On Sat, 24 Jun 2023 at 05:52, Matt Hoffner ***@***.***> wrote:
ctransformers just added support for this and it seems to be working. It's
been my goto for any model that doesn't work in llama
https://github.com/marella/ctransformers/tree/main
I've got some spaces setup based on their demo setup here:
https://huggingface.co/spaces/matthoffner/falcon-mini
—
Reply to this email directly, view it on GitHub
<#418 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/ABBMHMDGQNNZD3VIWMCJ2QTXMYXK3ANCNFSM6AAAAAAZRIZUFU>
.
You are receiving this because you authored the thread.Message ID:
***@***.***
com>
--
Thanks and regards,
Karrtik
|
Beta Was this translation helpful? Give feedback.
-
I wonder here too, the correct way seems to be to integrate: https://github.com/cmp-nct/ggllm.cpp I wonder if this is a target in the future? |
Beta Was this translation helpful? Give feedback.
-
As i understand for Falcon 40b, there is some change w.r.t ggml because of which this new fork has been created: https://github.com/cmp-nct/ggllm.cpp
I would like to know from the community if someone has tried using python bindings llama-cpp-python with falcon 40b GGML files from hugging face? https://huggingface.co/TheBloke/falcon-40b-instruct-GGML
If so, can you please share if any changes were needed to make this work?
Thanks in advance,
Beta Was this translation helpful? Give feedback.
All reactions