Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add qwen2-0.5B-instruct support. #35

Open
lavilao opened this issue Jul 3, 2024 · 5 comments
Open

Add qwen2-0.5B-instruct support. #35

lavilao opened this issue Jul 3, 2024 · 5 comments
Labels
enhancement New feature or request

Comments

@lavilao
Copy link

lavilao commented Jul 3, 2024

Its a really good model for its size and it aligns with the goal of this project.

@jncraton
Copy link
Owner

jncraton commented Jul 3, 2024

I've explored this model, and I hope to add it at some point. It isn't currently supported in the ctranslate2 backend that we use for inference. If/when it is supported there it shouldn't be difficult to add here.

@jncraton jncraton added the enhancement New feature or request label Jul 3, 2024
@lavilao
Copy link
Author

lavilao commented Jul 5, 2024

Umm, small question. Now that llama-cpp supports flan-t5, would you consider to change from ctranslate2 to it? it would allow a broader model and quantization support (making it easier to mantain as you dont have to convert your own models).
PD: support was added yesterday so llama-cpp-python support is still not there but should be comming.

@jncraton
Copy link
Owner

jncraton commented Jul 5, 2024

I'm open to that. At the moment I'm not aware of well-maintained Python bindings that support batched inference for llama-cpp. I would prefer not to lose that performance benefit. There is work being done in on this in llama-cpp-python.

@jncraton
Copy link
Owner

@lavilao There's still no Qwen2 (or 2.5) support, but I did recently update the package to support the following instruct models:

  • Llama-3.2-3B-Instruct
  • Llama-3.2-1B-Instruct
  • SmolLM-1.7B-Instruct
  • SmolLM-360M-Instruct
  • SmolLM-135M-Instruct
  • h2o-danube3-500m-chat
  • h2o-danube3-4b-chat

@lavilao
Copy link
Author

lavilao commented Sep 26, 2024

Awesome, i wonder if llama 3.2 1b Will run on My potato.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

2 participants