Skip to content

Conversation

csabakecskemeti
Copy link
Contributor

Nvidia uses the LLaMAForCausalLM string in their config.json so though there's support for LlamaForCausalLM
@Model.register("LlamaForCausalLM", "MistralForCausalLM", "MixtralForCausalLM")
the conversion failes on the case.

I've added the LLaMAForCausalLM

Example models with this arch string:

  • nvidia/Llama3-ChatQA-2-8B

  • nvidia/Llama3-ChatQA-2-70B

  • I have read the contributing guidelines

  • Self-reported review complexity:

    • Low
    • Medium
    • High

@github-actions github-actions bot added the python python script changes label Sep 14, 2024
@ggerganov ggerganov merged commit 3c7989f into ggml-org:master Sep 15, 2024
9 checks passed
dsx1986 pushed a commit to dsx1986/llama.cpp that referenced this pull request Oct 29, 2024
arthw pushed a commit to arthw/llama.cpp that referenced this pull request Nov 15, 2024
arthw pushed a commit to arthw/llama.cpp that referenced this pull request Nov 18, 2024
Nexesenex pushed a commit to Nexesenex/croco.cpp that referenced this pull request Feb 25, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

python python script changes

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants