Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

nvidia uses the LLaMAForCausalLM string in their config.json, example… #9485

Merged

Conversation

csabakecskemeti
Copy link
Contributor

Nvidia uses the LLaMAForCausalLM string in their config.json so though there's support for LlamaForCausalLM
@Model.register("LlamaForCausalLM", "MistralForCausalLM", "MixtralForCausalLM")
the conversion failes on the case.

I've added the LLaMAForCausalLM

Example models with this arch string:

  • nvidia/Llama3-ChatQA-2-8B

  • nvidia/Llama3-ChatQA-2-70B

  • I have read the contributing guidelines

  • Self-reported review complexity:

    • Low
    • Medium
    • High

@github-actions github-actions bot added the python python script changes label Sep 14, 2024
@ggerganov ggerganov merged commit 3c7989f into ggerganov:master Sep 15, 2024
9 checks passed
dsx1986 pushed a commit to dsx1986/llama.cpp that referenced this pull request Oct 29, 2024
arthw pushed a commit to arthw/llama.cpp that referenced this pull request Nov 15, 2024
arthw pushed a commit to arthw/llama.cpp that referenced this pull request Nov 18, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
python python script changes
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants